Goto

Collaborating Authors

 us ai safety institute


There are more than 120 AI bills in Congress right now

MIT Technology Review

A bill typically needs to pass a committee, or a smaller body of Congress, before it is voted on by the whole Congress. Many will fall short at this stage, while others will simply be introduced and then never spoken of again. This happens because there are so many bills presented in each session, and not all of them are given equal consideration. If the leaders of a party don't feel a bill from one of its members can pass, they may not even try to push it forward. And then, depending on the makeup of Congress, a bill's sponsor usually needs to get some members of the opposite party to support it for it to pass.


OpenAI and Anthropic agree to share their models with the US AI Safety Institute

Engadget

OpenAI and Anthropic have agreed to share AI models -- before and after release -- with the US AI Safety Institute. The agency, established through an executive order by President Biden in 2023, will offer safety feedback to the companies to improve their models. OpenAI CEO Sam Altman hinted at the agreement earlier this month. "Safety is essential to fueling breakthrough technological innovation. With these agreements in place, we look forward to beginning our technical collaborations with Anthropic and OpenAI to advance the science of AI safety," Elizabeth Kelly, director of the US AI Safety Institute, wrote in a statement.


OpenAI vows to provide the US government early access to its next AI model

Engadget

OpenAI will give the US AI Safety Institute early access to its next model as part of its safety efforts, Sam Altman has revealed in a tweet. Apparently, the company has been working with the consortium "to push forward the science of AI evaluations." The National Institute of Standards and Technology (NIST) has formally established the Artificial Intelligence Safety Institute earlier this year, though Vice President Kamala Harris announced it back in 2023 at the UK AI Safety Summit. Based on the NIST's description of the consortium, it's meant "to develop science-based and empirically backed guidelines and standards for AI measurement and policy, laying the foundation for AI safety across the world." The company, along with DeepMind, similarly pledged to share AI models with the UK government last year.